A Probabilistic Model for Learning Multi-Prototype Word Embeddings
نویسندگان
چکیده
pages 151–160, Dublin, Ireland, August 23-29 2014. A Probabilistic Model for Learning Multi-Prototype Word Embeddings Fei Tian†, Hanjun Dai∗, Jiang Bian‡, Bin Gao‡, Rui Zhang?, Enhong Chen†, Tie-Yan Liu‡ †University of Science and Technology of China, Hefei, P.R.China ∗Fudan University, Shanghai, P.R.China ‡Microsoft Research, Building 2, No. 5 Danling Street, Beijing, P.R.China ?Sun Yat-Sen University, Guangzhou, P.R.China †[email protected], †[email protected], ∗[email protected], ‡{jibian, bingao, tyliu}@microsoft.com, [email protected] Abstract
منابع مشابه
Improving Twitter Sentiment Classification Using Topic-Enriched Multi-Prototype Word Embeddings
It has been shown that learning distributed word representations is highly useful for Twitter sentiment classification. Most existing models rely on a single distributed representation for each word. This is problematic for sentiment classification because words are often polysemous and each word can contain different sentiment polarities under different topics. We address this issue by learnin...
متن کاملBridging Text and Knowledge by Learning Multi-Prototype Entity Mention Embedding
Integrating text and knowledge into a unified semantic space has attracted significant research interests recently. However, the ambiguity in the common space remains a challenge, namely that the same mention phrase usually refers to various entities. In this paper, to deal with the ambiguity of entity mentions, we propose a novel Multi-Prototype Mention Embedding model, which learns multiple s...
متن کاملBridge Text and Knowledge by Learning Multi-Prototype Entity Mention Embedding
Integrating text and knowledge into a unified semantic space has attracted significant research interests recently. However, the ambiguity in the common space remains a challenge, namely that the same mention phrase usually refers to various entities. In this paper, to deal with the ambiguity of entity mentions, we propose a novel Multi-Prototype Mention Embedding model, which learns multiple s...
متن کاملContext-Dependent Sense Embedding
Word embedding has been widely studied and proven helpful in solving many natural language processing tasks. However, the ambiguity of natural language is always a problem on learning high quality word embeddings. A possible solution is sense embedding which trains embedding for each sense of words instead of each word. Some recent work on sense embedding uses context clustering methods to dete...
متن کاملContext-Specific and Multi-Prototype Character Representations
Unsupervised word representations have demonstrated improvements in predictive generalization on various NLP tasks. Much effort has been devoted to effectively learning word embeddings, but little attention has been given to distributed character representations, although such character-level representations could be very useful for a variety of NLP applications in intrinsically “character-base...
متن کامل